Robust inversion, dimensionality reduction, and randomized sampling
نویسندگان
چکیده
We consider a class of inverse problems in which the forward model is the solution operator to linear ODEs or PDEs. This class admits several dimensionality-reduction techniques based on data averaging or sampling, which are especially useful for large-scale problems. We survey these approaches and their connection to stochastic optimization. The data-averaging approach is only viable, however, for a least-squares misfit, which is sensitive to outliers in the data and artifacts unexplained by the forward model. This motivates us to propose a robust formulation based on the Student’s t-distribution of the error. We demonstrate how the corresponding penalty function, together with the sampling approach, can obtain good results for a large-scale seismic inverse problem with 50% corrupted data.
منابع مشابه
Randomized Matrix Decompositions using R
The singular value decomposition (SVD) is among the most ubiquitous matrix factorizations. Specifically, it is a cornerstone algorithm for data analysis, dimensionality reduction and data compression. However, despite modern computer power, massive datasets pose a computational challenge for traditional SVD algorithms. We present the R package rsvd, which enables the fast computation of the SVD...
متن کاملProbabilistic inference of multi-Gaussian fields from indirect hydrological data using circulant embedding and dimensionality reduction
We present a Bayesian inversion method for the joint inference of high-dimensional multiGaussian hydraulic conductivity fields and associated geostatistical parameters from indirect hydrological data. We combine Gaussian process generation via circulant embedding to decouple the variogram from grid cell specific values, with dimensionality reduction by interpolation to enable Markov chain Monte...
متن کاملOptimal and scalable methods to approximate the solutions of large-scale Bayesian problems: Theory and application to atmospheric inversions and data assimilation
This paper provides a detailed theoretical analysis of methods to approximate the solutions of high-dimensional (> 10) linear Bayesian problems. An optimal low-rank projection that maximizes the information content of the Bayesian inversion is proposed and efficiently constructed using a scalable randomized SVD algorithm. Useful optimality results are established for the associated posterior er...
متن کاملLow-dimensional Embedding of Large-scale Infinite-dimensional Function Spaces with Application to Human Brain Connectome
Graph-based dimensionality reduction techniques assume that each datapoint can be written as a fixed width vector with a well-defined distance measure among datapoints; also, they typically assume that the number of instances is small enough to perform matrix inversion or pseudo-inversion. This paper considers dimensionality reduction on data using graph-based methods when two extreme circumsta...
متن کاملDimensionality Reduction for Sparse and Structured Matrices
Dimensionality reduction has become a critical tool for quickly solving massive matrix problems. Especially in modern data analysis and machine learning applications, an overabundance of data features or examples can make it impossible to apply standard algorithms efficiently. To address this issue, it is often possible to distill data to a much smaller set of informative features or examples, ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Math. Program.
دوره 134 شماره
صفحات -
تاریخ انتشار 2012